Query Selector–Efficient transformer with sparse attention
نویسندگان
چکیده
We present Query Selector - a sparse attention Transformer algorithm especially efficient for long-term time series forecasting. Time forecasting is an old and important area of statistical research with vast practical applications to solving real life problems. In recent years, there has been growing interest in applying Deep Learning algorithms modeling. However, the best performing algorithm, Transformer, shows some problems modeling long term due inherent weakness this solution i.e. self-attention mechanism. Memory requirements canonical architecture are quadratically sequence length. attempt address problem.
منابع مشابه
Transformer fault diagnosis using continuous sparse autoencoder.
This paper proposes a novel continuous sparse autoencoder (CSAE) which can be used in unsupervised feature learning. The CSAE adds Gaussian stochastic unit into activation function to extract features of nonlinear data. In this paper, CSAE is applied to solve the problem of transformer fault recognition. Firstly, based on dissolved gas analysis method, IEC three ratios are calculated by the con...
متن کاملA Selective Sparse Coding Model with Embedded Attention Mechanism
Sparse coding theory demonstrates that the neurons in the primary visual cortex form a sparse representation of natural scenes in the viewpoint of statistics, but a typical scene contains many different patterns (corresponding to neurons in cortex) competing for neural representation because of the limited processing capacity of the visual system. We propose an attention-guided sparse coding mo...
متن کاملQuery by Attention: Visually Searchable Information Maps
The Problem and Motivation: Vast amounts of useful information are accessible to us in textual databases, yet we prefer graphs, charts, and maps to interpret this information and make decisions. One advantage of these graphical presentations is that the user can query by attention – answer questions by controlling visual attention (and receiving immediate feedback), rather than manipulating a d...
متن کاملSalientDSO: Bringing Attention to Direct Sparse Odometry
Although cluttered indoor scenes have a lot of useful high-level semantic information which can be used for mapping and localization, most Visual Odometry (VO) algorithms rely on the usage of geometric features such as points, lines and planes. Lately, driven by this idea, the joint optimization of semantic labels and obtaining odometry has gained popularity in the robotics community. The joint...
متن کاملSelective attention integrates sparse and population codes
How visual information is represented in the brain is the crucial question that drives most research in the field, and answering it will mean significant theoretical and practical advances in the understanding of the brain. Proponents of sparse representations suggest that highly selective, specialized neurons, introduced as cardinal cells by Barlow [1], explicitly code each percept. Energy min...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Software impacts
سال: 2022
ISSN: ['2665-9638']
DOI: https://doi.org/10.1016/j.simpa.2021.100187